mDeBERTa is the multilingual version of DeBERTa, employing ELECTRA-style pretraining and gradient-disentangled embedding sharing technology, demonstrating excellent performance in cross-lingual tasks like XNLI
Large Language Model
Transformers Supports Multiple Languages